AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Three-stage pre-training

# Three-stage pre-training

Qwen3 1.7B Base
Apache-2.0
Qwen3-1.7B-Base is the latest generation of large language models in the Tongyi series, offering a range of dense models and mixture-of-experts (MoE) models. It has made significant improvements in training data, model architecture, and optimization techniques.
Large Language Model Transformers
Q
unsloth
7,444
2
Qwen3 0.6B Base Unsloth Bnb 4bit
Apache-2.0
Qwen3-0.6B-Base is the latest generation of large language models in the Tongyi series. It has a parameter scale of 0.6B, supports 119 languages, and has a context length of up to 32,768 tokens.
Large Language Model Transformers
Q
unsloth
10.84k
1
Qwen3 0.6B Base
Apache-2.0
Qwen3-0.6B-Base is the latest generation of large language models in the Tongyi Qianwen series, offering a range of dense models and Mixture of Experts (MoE) models.
Large Language Model Transformers
Q
unsloth
10.84k
2
Qwen3 1.7B Base
Apache-2.0
Qwen3-1.7B is the latest 1.7 billion parameter base language model in the Qwen series, featuring a three-stage pre-training system and supporting 32k context length.
Large Language Model Transformers
Q
Qwen
19.24k
19
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase